7 research outputs found
Optimal Eigenvalue Approximation via Sketching
Given a symmetric matrix , we show from the simple sketch , where
is a Gaussian matrix with rows, that there is a
procedure for approximating all eigenvalues of simultaneously to within
additive error with large probability. Unlike the work of
(Andoni, Nguyen, SODA, 2013), we do not require that is positive
semidefinite and therefore we can recover sign information about the spectrum
as well. Our result also significantly improves upon the sketching dimension of
recent work for this problem (Needell, Swartworth, Woodruff FOCS 2022), and in
fact gives optimal sketching dimension. Our proof develops new properties of
singular values of for a Gaussian matrix and an matrix which may be of independent interest. Additionally we achieve
tight bounds in terms of matrix-vector queries. Our sketch can be computed
using matrix-vector multiplies, and by improving on lower
bounds for the so-called rank estimation problem, we show that this number is
optimal even for adaptive matrix-vector queries
Testing Positive Semidefiniteness Using Linear Measurements
We study the problem of testing whether a symmetric input matrix
is symmetric positive semidefinite (PSD), or is -far from the PSD
cone, meaning that , where
is the Schatten- norm of . In applications one often needs to quickly
tell if an input matrix is PSD, and a small distance from the PSD cone may be
tolerable. We consider two well-studied query models for measuring efficiency,
namely, the matrix-vector and vector-matrix-vector query models. We first
consider one-sided testers, which are testers that correctly classify any PSD
input, but may fail on a non-PSD input with a tiny failure probability. Up to
logarithmic factors, in the matrix-vector query model we show a tight
bound, while in the
vector-matrix-vector query model we show a tight
bound, for every . We also
show a strong separation between one-sided and two-sided testers in the
vector-matrix-vector model, where a two-sided tester can fail on both PSD and
non-PSD inputs with a tiny failure probability. In particular, for the
important case of the Frobenius norm, we show that any one-sided tester
requires queries. However we introduce
a bilinear sketch for two-sided testing from which we construct a Frobenius
norm tester achieving the optimal queries. We
also give a number of additional separations between adaptive and non-adaptive
testers. Our techniques have implications beyond testing, providing new methods
to approximate the spectrum of a matrix with Frobenius norm error using
dimensionality reduction in a way that preserves the signs of eigenvalues
Testing Hereditary Properties of Sequences
A hereditary property of a sequence is one that is preserved when restricting to subsequences. We show that there exist hereditary properties of sequences that cannot be tested with sublinear queries, resolving an open question posed by Newman et al. This proof relies crucially on an infinite alphabet, however; for finite alphabets, we observe that any hereditary property can be tested with a constant number of queries
Recommended from our members
Efficient Algorithms for Linear Regression and Spectrum Estimation
In this thesis we study efficient algorithms for solving very large linear algebra problems. We firstconsider the Kaczmarz method for solving linear systems, and develop a variant that is robust to a
small number of large corruptions, while still requiring only a small working memory. We provide
both theoretical guarantees for certain data distributions as well as empirical results showing that
our approach works well in practice. We then turn our attention to problems of quickly learning
spectral information about a matrix. The first such problem is PSD-testing where we give optimal
query complexity bounds (with respect to types of types of queries) for distinguishing between a
matrix being positive semi-definite versus having a large negative eigenvalue. Building on part of
this work, we then develop optimal sketches for learning the entire spectrum of a matrix to within
additive error. Finally we return our attention to solving linear systems and give new algorithms
that achieve optimal communication complexity for solving least-squares regression problems